7 research outputs found
HINT: Hierarchical Invertible Neural Transport for Density Estimation and Bayesian Inference
A large proportion of recent invertible neural architectures is based on a
coupling block design. It operates by dividing incoming variables into two
sub-spaces, one of which parameterizes an easily invertible (usually affine)
transformation that is applied to the other. While the Jacobian of such a
transformation is triangular, it is very sparse and thus may lack
expressiveness. This work presents a simple remedy by noting that (affine)
coupling can be repeated recursively within the resulting sub-spaces, leading
to an efficiently invertible block with dense triangular Jacobian. By
formulating our recursive coupling scheme via a hierarchical architecture, HINT
allows sampling from a joint distribution p(y,x) and the corresponding
posterior p(x|y) using a single invertible network. We demonstrate the power of
our method for density estimation and Bayesian inference on a novel data set of
2D shapes in Fourier parameterization, which enables consistent visualization
of samples for different dimensionalities
A Stein variational Newton method
Stein variational gradient descent (SVGD) was recently proposed as a general
purpose nonparametric variational inference algorithm [Liu & Wang, NIPS 2016]:
it minimizes the Kullback-Leibler divergence between the target distribution
and its approximation by implementing a form of functional gradient descent on
a reproducing kernel Hilbert space. In this paper, we accelerate and generalize
the SVGD algorithm by including second-order information, thereby approximating
a Newton-like iteration in function space. We also show how second-order
information can lead to more effective choices of kernel. We observe
significant computational gains over the original SVGD algorithm in multiple
test cases.Comment: 18 pages, 7 figure
Continuous Level Monte Carlo and Sample-Adaptive Model Hierarchies
This is the final version. Available from SIAM via the DOI in this record.In this paper, we present a generalisation of the Multilevel Monte Carlo (MLMC) method to a setting where the level parameter is a continuous variable. This Continuous Level Monte Carlo (CLMC) estimator provides a natural framework in PDE applications to adapt the model hierarchy to each sample. In addition, it can be made unbiased with respect to the expected value of the true quantity of interest provided the quantity of interest converges sufficiently fast. The practical implementation of the CLMC estimator is based on interpolating actual evaluations of the quantity of interest at a finite number of resolutions. As our new level parameter, we use the logarithm of a goal-oriented finite element error estimator for the accuracy of the quantity of interest. We prove the unbiasedness, as well as a complexity theorem that shows the same rate of complexity for CLMC as for MLMC. Finally, we provide some numerical evidence to support our theoretical results, by successfully testing CLMC on a standard PDE test problem. The numerical experiments demonstrate clear gains for sample-wise adaptive refinement strategies over uniform refinements